188 research outputs found

    Restructured eigenfilter matching for novelty detection in random textures

    Get PDF
    A new eigenfilter-based novelty detection approach to find abnormalities in random textures is presented. The proposed algorithm reconstructs a given texture twice using a subset of its own eigenfilter bank and a subset of a reference (template) eigenfilter bank, and measures the reconstruction error as the level of novelty. We then present an improved reconstruction generated by structurally matched eigenfilters through rotation, negation, and mirroring. We apply the method to the detection of defects in textured ceramic tiles. The method is over 90 % accurate, and is fast and amenable to implementation on a production line.

    Automatic Bootstrapping and Tracking of Object Contours

    Get PDF
    This work introduces a new fully automatic object tracking and segmentation framework. The framework consists of a motion based bootstrapping algorithm concurrent to a shape based active contour. The shape based active contour uses a finite shape memory that is automatically and continuously built from both the bootstrap process and the active contour object tracker. A scheme is proposed to ensure the finite shape memory is continuously updated but forgets unnecessary information. Two new ways of automatically extracting shape information from image data given a region of interest are also proposed. Results demonstrate that the bootstrapping stage provides important motion and shape information to the object tracker

    TEXEMS: Texture Exemplars for Defect Detection on Random Textured Surfaces

    Get PDF

    RAGS: Region-Aided Geometric Snake

    Get PDF
    Abstract—An enhanced, region-aided, geometric active contour that is more tolerant toward weak edges and noise in images is introduced. The proposed method integrates gradient flow forces with region constraints, composed of image region vector flow forces obtained through the diffusion of the region segmentation map. We refer to this as the Region-aided Geometric Snake or RAGS. The diffused region forces can be generated from any reliable region segmentation technique, greylevel or color. This extra region force gives the snake a global complementary view of the boundary information within the image which, along with the local gradient flow, helps detect fuzzy boundaries and overcome noisy regions. The partial differential equation (PDE) resulting from this integration of image gradient flow and diffused region flow is implemented using a level set approach. We present various examples and also evaluate and compare the performance of RAGS on weak boundaries and noisy images. Index Terms—Color snakes, deformable contours, geometric snakes, region segmentation, region-aided snakes, weak-edge leakage. I

    Video-SwinUNet: Spatio-temporal Deep Learning Framework for VFSS Instance Segmentation

    Full text link
    This paper presents a deep learning framework for medical video segmentation. Convolution neural network (CNN) and transformer-based methods have achieved great milestones in medical image segmentation tasks due to their incredible semantic feature encoding and global information comprehension abilities. However, most existing approaches ignore a salient aspect of medical video data - the temporal dimension. Our proposed framework explicitly extracts features from neighbouring frames across the temporal dimension and incorporates them with a temporal feature blender, which then tokenises the high-level spatio-temporal feature to form a strong global feature encoded via a Swin Transformer. The final segmentation results are produced via a UNet-like encoder-decoder architecture. Our model outperforms other approaches by a significant margin and improves the segmentation benchmarks on the VFSS2022 dataset, achieving a dice coefficient of 0.8986 and 0.8186 for the two datasets tested. Our studies also show the efficacy of the temporal feature blending scheme and cross-dataset transferability of learned capabilities. Code and models are fully available at https://github.com/SimonZeng7108/Video-SwinUNet

    Deep Compact Person Re-Identification with Distractor Synthesis via Guided DC-GANs

    Get PDF
    We present a dual-stream CNN that learns both appearance and facial features in tandem from still images and, after feature fusion, infers person identities. We then describe an alternative architecture of a single, lightweight ID-CondenseNet where a face detector-guided DC-GAN is used to generate distractor person images for enhanced training. For evaluation, we test both architectures on FLIMA, a new extension of an existing person re-identification dataset with added frame-by-frame annotations of face presence. Although the dual-stream CNN can outperform the CondenseNet approach on FLIMA, we show that the latter surpasses all state-of-the-art architectures in top-1 ranking performance when applied to the largest existing person re-identification dataset, MSMT17. We conclude that whilst re-identification performance is highly sensitive to the structure of datasets, distractor augmentation and network compression have a role to play for enhancing performance characteristics for larger scale applications

    Semantically selective augmentation for deep compact person re-identification

    Get PDF
    We present a deep person re-identification approach that combines semantically selective, deep data augmentation with clustering-based network compression to generate high performance, light and fast inference networks. In particular, we propose to augment limited training data via sampling from a deep convolutional generative adversarial network (DCGAN), whose discriminator is constrained by a semantic classifier to explicitly control the domain specificity of the generation process. Thereby, we encode information in the classifier network which can be utilized to steer adversarial synthesis, and which fuels our CondenseNet ID-network training. We provide a quantitative and qualitative analysis of the approach and its variants on a number of datasets, obtaining results that outperform the state-of-the-art on the LIMA dataset for long-term monitoring in indoor living spaces

    Tracking with active contours using dynamically updated shape information

    Get PDF
    An active contour based tracking framework is described that generates and integrates dynamic shape information without having to learn a priori shape constraints. This dynamic shape information is combined with dynamic photometric foreground model matching and background mismatching. Boundary based optical flow is also used to estimate the location of the object in each new frame, incorporating Procrustes shape alignment. Promising results under complex deformations of shape, varied levels of noise, and closeto-complete occlusion in complex textured backgrounds are presented.
    • …
    corecore